首页> 外文OA文献 >Efficient algorithms for training the parameters of hidden Markov models using stochastic expectation maximization EM training and Viterbi training
【2h】

Efficient algorithms for training the parameters of hidden Markov models using stochastic expectation maximization EM training and Viterbi training

机译:用于训练隐马尔可夫模型参数的高效算法   使用随机期望最大化Em培训和Viterbi培训

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Background: Hidden Markov models are widely employed by numerousbioinformatics programs used today. Applications range widely from comparativegene prediction to time-series analyses of micro-array data. The parameters ofthe underlying models need to be adjusted for specific data sets, for examplethe genome of a particular species, in order to maximize the predictionaccuracy. Computationally efficient algorithms for parameter training are thuskey to maximizing the usability of a wide range of bioinformatics applications. Results: We introduce two computationally efficient training algorithms, onefor Viterbi training and one for stochastic expectation maximization (EM)training, which render the memory requirements independent of the sequencelength. Unlike the existing algorithms for Viterbi and stochastic EM trainingwhich require a two-step procedure, our two new algorithms require only onestep and scan the input sequence in only one direction. We also implement thesetwo new algorithms and the already published linear-memory algorithm for EMtraining into the hidden Markov model compiler HMM-Converter and examine theirrespective practical merits for three small example models. Conclusions: Bioinformatics applications employing hidden Markov models canuse the two algorithms in order to make Viterbi training and stochastic EMtraining more computationally efficient. Using these algorithms, parametertraining can thus be attempted for more complex models and longer trainingsequences. The two new algorithms have the added advantage of being easier toimplement than the corresponding default algorithms for Viterbi training andstochastic EM training.
机译:背景:隐马尔可夫模型已被当今使用的众多生物信息学程序广泛采用。应用范围广泛,从比较基因预测到微阵列数据的时间序列分析。需要针对特定​​数据集(例如特定物种的基因组)调整基础模型的参数,以使预测准确性最大化。因此,用于参数训练的计算有效算法是最大化各种生物信息学应用程序可用性的关键。结果:我们引入了两种计算有效的训练算法,一种用于维特比训练,一种用于随机期望最大化(EM)训练,这使得内存需求与序列长度无关。与现有的用于维特比和随机EM训练的算法需要两步过程不同,我们的两种新算法仅需一步就可以在一个方向上扫描输入序列。我们还实现了这两种新算法以及已经发布的线性记忆算法,用于将EMtraining到隐藏的马尔可夫模型编译器HMM-Converter中,并针对三个小示例模型检验了它们各自的实用性。结论:采用隐马尔可夫模型的生物信息学应用程序可以使用这两种算法,以使Viterbi训练和随机EMtraining的计算效率更高。因此,使用这些算法,可以针对更复杂的模型和更长的训练序列尝试进行参数训练。与维特比训练和随机EM训练的相应默认算法相比,这两种新算法的优点是易于实现。

著录项

代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号